Classification of Musical Timbre Using Bayesian Networks

نویسندگان

  • Patrick J. Donnelly
  • John W. Sheppard
چکیده

In this article, we explore the use of Bayesian networks for identifying the timbre of musical instruments. Peak spectral amplitude in ten frequency windows is extracted for each of 20 time windows to be used as features. Over a large data set of 24,000 audio examples covering the full musical range of 24 different common orchestral instruments, four different Bayesian network structures, including naive Bayes, are examined and compared with two support vector machines and a k-nearest neighbor classifier. Classification accuracy is examined by instrument, instrument family, and data set size. Bayesian networks with conditional dependencies in the time and frequency dimensions achieved 98 percent accuracy in the instrument classification task and 97 percent accuracy in the instrument family identification task. These results demonstrate a significant improvement over the previous approaches in the literature on this data set. Additionally, we tested our Bayesian approach on thewidely used Iowamusical instrument data set, with similar results. The identification of musical instruments in audio recordings is a frequently explored, yet unsolved, machine learning problem. Despite a number of experiments in the literature over the years, no single feature-extraction scheme or learning approach has emerged as a definitive solution to this classification problem. The ability of a computer to learn to identify musical instruments is an important problem within the field of music information retrieval, with high commercial value. For instance, companies could automatically index their music libraries based on the musical instruments present in the recording, allowing search and retrieval by specific musical instrument. Timbre identification is also important to the tasks of musical genre categorization, automatic score creation, and track separation. This work investigates classification of single, monophonic musical instruments using several different Bayesian network structures and a featureextraction scheme based on a psychoacoustic definition of timbre. The results of this seminal use of graphical models in the task of musical instrument classification are compared with the baseline algorithms of support vector machines and a k-nearest neighbor classifier. Computer Music Journal, 37:4, pp. 70–86, Winter 2014 doi:10.1162/COMJ a 00210 c © 2014 Massachusetts Institute of Technology. Timbre When a musical instrument plays a note, we perceive a musical pitch, the instrument playing that note, and other aspects, like loudness. Timbre, or tone color, is the psychoacoustic property of sound that allows the human brain to readily distinguish between two instances of the same note, each played on a different instruments. The primary musical pitch we perceive is usually the first harmonic partial, known as the fundamental frequency. Pitched instruments are those whose partials are approximate integer multiples of the fundamental frequency. With the exception of unpitched percussion, orchestral instruments are pitched. The perception of timbre depends on the presence of harmonics (i.e., spectrum), as well as the fine timing (envelope) of each harmonic constituent (partial) of the musical signal (Donnelly and Limb 2009).

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

BAYESIAN APPROACHES TO MUSICAL INSTRUMENT CLASSIFICATION USING TIMBRE SEGMENTATION by

The task of identifying musical instruments in an audio recording is a difficult problem. While there exists a body of literature on single instrument identification, little research has been performed on the more complex, but real-world, situation of more than one instrument present in the signal. This work proposes a Bayesian method for multi-label classification of musical instrument timbre....

متن کامل

Radial/elliptical Basis Function Neural Networks for Timbre Classification

This paper outlines a RBF/EBF neural network approach for automatic musical instrument classification using salient feature extraction techniques with a combination of supervised and unsupervised learning schemes. 829 monophonic sound examples (86% Siedlaczek Library [2], 14% other sources) from the string, brass, and woodwind families with a variety of performance techniques, dynamics, and pit...

متن کامل

Binary Decision Tree Classification of Musical Sounds

This paper presents a novel method of classifying musical sounds. An earlier work has shown the ability of a subset of the timbre attributes of musical sounds to classify musical sounds correctly in instrument families. This work focuses on the interpretation of the timbre attributes. The question is: which timbre attributes are useful for the classification of the sounds? These attributes are ...

متن کامل

Classification of Iranian traditional musical modes (DASTGÄH) with artificial neural network

The concept of Iranian traditional musical modes, namely DASTGÄH, is the basis for the traditional music system. The concept introduces seven DASTGÄHs. It is not an easy process to distinguish these modes and such practice is commonly performed by an experienced person in this field. Apparently, applying artificial intelligence to do such classification requires a combination of the basic infor...

متن کامل

Musical Instrument Extraction through Timbre Classification

Contemporary technological advancement of internet and online servers allows many musical pieces to be readily available to the users to enjoy. The users may listen to the music, share with friends, or create another musical piece by either remixing or sampling. One may desire to simply play the music as it is or sample just one instrument out of the music, however, this task can be challenging...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Computer Music Journal

دوره 37  شماره 

صفحات  -

تاریخ انتشار 2013